Stability and Generalization of Learning Algorithms that Converge to Global Optima

نویسندگان

  • Zachary B. Charles
  • Dimitris S. Papailiopoulos
چکیده

We establish novel generalization bounds for learning algorithms that converge to global minima. We do so by deriving black-box stability results that only depend on the convergence of a learning algorithm and the geometry around the minimizers of the loss function. The results are shown for nonconvex loss functions satisfying the Polyak-Łojasiewicz (PL) and the quadratic growth (QG) conditions. We further show that these conditions arise for some neural networks with linear activations. We use our black-box results to establish the stability of optimization algorithms such as stochastic gradient descent (SGD), gradient descent (GD), randomized coordinate descent (RCD), and the stochastic variance reduced gradient method (SVRG), in both the PL and the strongly convex setting. Our results match or improve state-of-the-art generalization bounds and can easily be extended to similar optimization algorithms. Finally, we show that although our results imply comparable stability for SGD and GD in the PL setting, there exist simple neural networks with multiple local minima where SGD is stable but GD is not.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Differential Evolution and Spatial Distribution based Local Search for Training Fuzzy Wavelet Neural Network

Abstract   Many parameter-tuning algorithms have been proposed for training Fuzzy Wavelet Neural Networks (FWNNs). Absence of appropriate structure, convergence to local optima and low speed in learning algorithms are deficiencies of FWNNs in previous studies. In this paper, a Memetic Algorithm (MA) is introduced to train FWNN for addressing aforementioned learning lacks. Differential Evolution...

متن کامل

Tabu-KM: A Hybrid Clustering Algorithm Based on Tabu Search Approach

  The clustering problem under the criterion of minimum sum of squares is a non-convex and non-linear program, which possesses many locally optimal values, resulting that its solution often falls into these trap and therefore cannot converge to global optima solution. In this paper, an efficient hybrid optimization algorithm is developed for solving this problem, called Tabu-KM. It gathers the ...

متن کامل

A Meta-heuristic Algorithm for Global Numerical Optimization Problems inspired by Vortex in fluid physics

One of the most important issues in engineering is to find the optimal global points of the functions used. It is not easy to find such a point in some functions due to the reasons such as large number of dimensions or inability to derive them from the function. Also in engineering modeling, we do not have the relationships of many functions, but we can input and output them as a black box. The...

متن کامل

An Improved Particle Swarm Optimizer Based on a Novel Class of Fast and Efficient Learning Factors Strategies

The particle swarm optimizer (PSO) is a population-based metaheuristic optimization method that can be applied to a wide range of problems but it has the drawbacks like it easily falls into local optima and suffers from slow convergence in the later stages. In order to solve these problems, improved PSO (IPSO) variants, have been proposed. To bring about a balance between the exploration and ex...

متن کامل

A Hybrid Algorithm using Firefly, Genetic, and Local Search Algorithms

In this paper, a hybrid multi-objective algorithm consisting of features of genetic and firefly algorithms is presented. The algorithm starts with a set of fireflies (particles) that are randomly distributed in the solution space; these particles converge to the optimal solution of the problem during the evolutionary stages. Then, a local search plan is presented and implemented for searching s...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1710.08402  شماره 

صفحات  -

تاریخ انتشار 2017